Learning Imprecise Hidden Markov Models

نویسندگان

  • Arthur Van Camp
  • Gert de Cooman
  • Jasper De Bock
  • Erik Quaeghebeur
  • Filip Hermans
چکیده

Consider a stationary precise hidden Markov model (HMM) with n hidden states Xk, taking values xk in a set {1, . . . ,m} and n observations Ok, taking values ok. Both the marginal model pX1(x1), the emission models pOk|Xk(ok|xk) and the transition models pXk|Xk−1(xk|xk−1) are unknown. We can then use the Baum–Welch algorithm [see, e.g., 4] to get a maximum-likelihood estimate of these models. The Baum–Welch algorithm constructs the expected number of transitions ni j := ∑k=2 pXk−1,Xk|O1,...,On(i, j|o1, . . . ,on) from state i to state j in the whole Markov chain. If we do not have enough data to justify a precise probability model, such as the one we obtain using the classical Baum–Welch algorithm, then the approach we present is useful. Our contribution exists of a method for learning imprecise transition probabilities in an HMM. We are not aware of another such method in the literature. These transitions from a state Xk−1 = i to a state Xk = j are multinomial processes. The imprecise Dirichlet model (IDM) is a convenient model for describing uncertainty about such processes [3]. In order to learn using an IDM, we need the number of transitions and a choice for the pseudocounts s. Since the hidden states are unavailable, our method consists in taking the expected1 number of transitions (positive real numbers instead of natural numbers), derived from the Baum–Welch algorithm, rather than real counts. So, the lower probability for state j conditional on state i is estimated by Q({ j}|i) := ni j/(s+∑j=1 ni j). Learning an imprecise HMM, like our method does, is necessary before being able to make imprecise inferences. There are algorithms, like MePICTIr [2] and EstiHMM [1], to make exact inferences in an imprecise HMM. We applied our method to the following application. Given an observation sequence of counted earthquakes in a number of consecutive years, we are interested in predicting the number of earthquakes in the next years. The observed number of earthquakes in a year k is assumed to be generated by a Poisson process with rate λXk . Both the emission probabilities and the marginal probability are kept precise, while the transition probabilities are made imprecise, using our method. The following figure shows the credal sets for the transition model in the case m = 3, for different pseudocounts s. λ2 λ1 λ3 s = 1 s = 2 s = 3 s = 5 s = 10 s = 20 s = 50

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Partially-Hidden Markov Models

This paper addresses the problem of Hidden Markov Models (HMM) training and inference when the training data are composed of feature vectors plus uncertain and imprecise labels. The “soft” labels represent partial knowledge about the possible states at each time step and the “softness” is encoded by belief functions. For the obtained model, called a Partially-Hidden Markov Model (PHMM), the tra...

متن کامل

Robust classification of multivariate time series by imprecise hidden Markov models

A novel technique to classify time series with imprecise hidden Markov models is presented. The learning of these models is achieved by coupling the EM algorithm with the imprecise Dirichlet model. In the stationarity limit, each model corresponds to an imprecise mixture of Gaussian densities, this reducing the problem to the classification of static, imprecise-probabilistic, information. Two c...

متن کامل

EstiHMM: an efficient algorithm for state sequence prediction in imprecise hidden Markov models

We develop an efficient algorithm that calculates the maximal state sequences in an imprecise hidden Markov model by means of coherent lower previsions. Initial results show that this algorithm is able to robustify the inferences made by a classical precise model. Keywords— Imprecise hidden Markov model, coherent lower prevision, epistemic irrelevance, maximal state sequence, Viterbi algorithm.

متن کامل

Training Generalized Hidden Markov Model With Interval Probability Parameters

Recently generalized interval probability was proposed as a new mathematical formalism of imprecise probability. It provides a simplified probabilistic calculus based on its definitions of conditional probability and independence. The Markov property can be described in a form similar to classical probability. In this paper, an expectation-maximization approach is developed to train generalized...

متن کامل

Introducing Busy Customer Portfolio Using Hidden Markov Model

Due to the effective role of Markov models in customer relationship management (CRM), there is a lack of comprehensive literature review which contains all related literatures. In this paper the focus is on academic databases to find all the articles that had been published in 2011 and earlier. One hundred articles were identified and reviewed to find direct relevance for applying Markov models...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2011